gradient descent - définition. Qu'est-ce que gradient descent
Diclib.com
Dictionnaire ChatGPT
Entrez un mot ou une phrase dans n'importe quelle langue 👆
Langue:

Traduction et analyse de mots par intelligence artificielle ChatGPT

Sur cette page, vous pouvez obtenir une analyse détaillée d'un mot ou d'une phrase, réalisée à l'aide de la meilleure technologie d'intelligence artificielle à ce jour:

  • comment le mot est utilisé
  • fréquence d'utilisation
  • il est utilisé plus souvent dans le discours oral ou écrit
  • options de traduction de mots
  • exemples d'utilisation (plusieurs phrases avec traduction)
  • étymologie

Qu'est-ce (qui) est gradient descent - définition


Gradient descent         
  • An animation showing the first 83 iterations of gradient descent applied to this example. Surfaces are [[isosurface]]s of <math>F(\mathbf{x}^{(n)})</math> at current guess <math>\mathbf{x}^{(n)}</math>, and arrows show the direction of descent. Due to a small and constant step size, the convergence is slow.
  • Illustration of gradient descent on a series of [[level set]]s
  • Fog in the mountains
  • The steepest descent algorithm applied to the [[Wiener filter]]<ref>Haykin, Simon S. Adaptive filter theory. Pearson Education India, 2008. - p. 108-142, 217-242</ref>
OPTIMIZATION ALGORITHM
Steepest descent; Gradient ascent; Gradient descent method; Steepest ascent; Gradient Descent; Gradient descent optimization; Gradient-based optimization; Gradient descent with momentum
In mathematics, gradient descent (also often called steepest descent) is a first-order iterative optimization algorithm for finding a local minimum of a differentiable function. The idea is to take repeated steps in the opposite direction of the gradient (or approximate gradient) of the function at the current point, because this is the direction of steepest descent.
Stochastic gradient descent         
OPTIMIZATION ALGORITHM
Gradient descent in machine learning; Incremental gradient descent; AdaGrad; Adagrad; Adam (optimization algorithm); Momentum (machine learning); RMSProp; Applications of stochastic gradient descent; Stochastic Gradient Descent; SGD optimizer; Adam optimizer; RMSprop
Stochastic gradient descent (often abbreviated SGD) is an iterative method for optimizing an objective function with suitable smoothness properties (e.g.
Gradient         
  • ''f''(''x'',''y'') {{=}} −(cos<sup>2</sup>''x'' + cos<sup>2</sup>''y'')<sup>2</sup>}} depicted as a projected [[vector field]] on the bottom plane.
  • 1=''f''(''x'', ''y'') = ''xe''<sup>−(''x''<sup>2</sup> + ''y''<sup>2</sup>)</sup>}} is plotted as arrows over the pseudocolor plot of the function.
MULTI-VARIABLE GENERALIZATION OF THE DERIVATIVE
Gradient vector; Gradients; Gradient (calculus); Gradient of a scalar; Gradient Operator; Grad operator
·adj Moving by steps; walking; as, gradient automata.
II. Gradient ·adj Adapted for walking, as the feet of certain birds.
III. Gradient ·noun The rate of regular or graded ascent or descent in a road; grade.
IV. Gradient ·noun A part of a road which slopes upward or downward; a portion of a way not level; a grade.
V. Gradient ·adj Rising or descending by regular degrees of inclination; as, the gradient line of a railroad.
VI. Gradient ·noun The rate of increase or decrease of a variable magnitude, or the curve which represents it; as, a thermometric gradient.